Periodic Stepsize Adaptation
نویسندگان
چکیده
Previously, Bottou and LeCun [1] established that the second-order stochastic gradient descent (SGD) method can potentially achieve generalization performance as well as empirical optimum in a single pass through the training examples. However, second-order SGD requires computing the inverse of the Hessian matrix of the loss function, which is usually prohibitively expensive. Recently, we invented a new second-order SGD method, called Periodic Stepsize Adaptation (PSA). PSA explores a simple linear relation between the Hessian matrix and the Jacobian matrix of the mapping function. Instead of approximating Hessian, PSA approximates the Jacobian matrix which is proved to be simpler and more effective than approximating Hessian in an on-line setting. Experimental results for conditional random fields (CRF) and neural networks (NN) show that single-pass performance of PSA is very close to empirical optimum.
منابع مشابه
L4: Practical loss-based stepsize adaptation for deep learning
We propose a stepsize adaptation scheme for stochastic gradient descent. It operates directly with the loss function and rescales the gradient in order to make fixed predicted progress on the loss. We demonstrate its capabilities by strongly improving the performance of Adam and Momentum optimizers. The enhanced optimizers with default hyperparameters consistently outperform their constant step...
متن کاملNew Proposed Second-Order ASDM Using OTAs
In this work, we present a new proposal for the second-order Adaptive Sigma Delta Modulation (ASDM). The new proposed Adaptation scheme is based on using Operational Transconductance Amplifier (OTAs) as an Integrator and as an Amplifier to adapt the quantizer stepsize to control the voltage gain by feedback the quantizer output through adaptation scheme. The stepsize is changing up or down by t...
متن کاملRecursive Adaptation of Stepsize Parameter for Non-stationary Environments
In this article, we propose a method to adapt stepsize parameters used in reinforcement learning for dynamic environments. In general reinforcement learning situations, a stepsize parameter is decreased to zero during learning, because the environment is generally supposed to be noisy but stationary, such that the true expected rewards are fixed. On the other hand, we assume that in the real wo...
متن کاملOn the periodic solutions of discrete Hamiltonian systems
Almost all numerical methods for solving conservative problems cannot avoid a more or less perceptible drift phenomenon. Considering that the drift would be absent on a periodic or quasi-periodic solution, one way to eliminate such unpleasant phenomenon is to look for discrete periodic or quasi-periodic solutions. It is quite easy to show that only symmetric methods are able to provide solution...
متن کاملAn adaptive homotopy approach for non-selfadjoint eigenvalue problems
This paper presents adaptive algorithms for eigenvalue problems associated with non-selfadjoint partial differential operators. The basis for the developed algorithms is a homotopy method which departs from a wellunderstood selfadjoint problem. Apart from the adaptive grid refinement, the progress of the homotopy as well as the solution of the iterative method are adapted to balance the contrib...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2008